Reusing Monolingual Pre-Trained Models by Cross-Connecting Seq2seq Models for Machine Translation
نویسندگان
چکیده
This work uses sequence-to-sequence (seq2seq) models pre-trained on monolingual corpora for machine translation. We pre-train two seq2seq with the source and target languages, then combine encoder of language model decoder model, i.e., cross-connection. add an intermediate layer between to help mapping each other since modules are completely independently. These can as a multilingual because one be cross-connected another any language, while their capacity is not affected by number languages. will demonstrate that our method improves translation performance significantly over random baseline. Moreover, we analyze appropriate choice layer, importance part change along size bitext.
منابع مشابه
Domain Adaptation of Statistical Machine Translation Models with Monolingual Data for Cross Lingual Information Retrieval
Statistical Machine Translation (SMT) is often used as a black-box in CLIR tasks. We propose an adaptation method for an SMT model relying on the monolingual statistics that can be extracted from the document collection (both source and target if available). We evaluate our approach on CLEF Domain Specific task (German-English and English-German) and show that very simple document collection st...
متن کاملImproving Neural Machine Translation Models with Monolingual Data
Neural Machine Translation (NMT) has obtained state-of-the art performance for several language pairs, while only using parallel data for training. Monolingual data plays an important role in boosting fluency for phrase-based statistical machine translation, and we investigate the use of monolingual data for neural machine translation (NMT). In contrast to previous work, which integrates a sepa...
متن کاملLearning Translation Models from Monolingual Continuous Representations
Translation models often fail to generate good translations for infrequent words or phrases. Previous work attacked this problem by inducing new translation rules from monolingual data with a semi-supervised algorithm. However, this approach does not scale very well since it is very computationally expensive to generate new translation rules for only a few thousand sentences. We propose a much ...
متن کاملSentence Simplification by Monolingual Machine Translation
In this paper we describe a method for simplifying sentences using Phrase Based Machine Translation, augmented with a re-ranking heuristic based on dissimilarity, and trained on a monolingual parallel corpus. We compare our system to a word-substitution baseline and two state-of-the-art systems, all trained and tested on paired sentences from the English part of Wikipedia and Simple Wikipedia. ...
متن کاملLog-linear Combinations of Monolingual and Bilingual Neural Machine Translation Models for Automatic Post-Editing
This paper describes the submission of the AMU (Adam Mickiewicz University) team to the Automatic Post-Editing (APE) task of WMT 2016. We explore the application of neural translation models to the APE problem and achieve good results by treating different models as components in a log-linear model, allowing for multiple inputs (the MT-output and the source) that are decoded to the same target ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Applied sciences
سال: 2021
ISSN: ['2076-3417']
DOI: https://doi.org/10.3390/app11188737